{37 () Bayesian Network Classiiers. *
نویسندگان
چکیده
Recent work in supervised learning has shown that a surprisingly simple Bayesian classiier with strong assumptions of independence among features, called naive Bayes, is competitive with state-of-the-art classiiers such as C4.5. This fact raises the question of whether a classiier with less restrictive assumptions can perform even better. In this paper we evaluate approaches for inducing classiiers from data, based on the theory of learning Bayesian networks. These networks are factored representations of probability distributions that generalize the naive Bayesian classiier and explicitly represent statements about independence. Among these approaches we single out a method we call Tree Augmented Naive Bayes (TAN), which outperforms naive Bayes, yet at the same time maintains the computational simplicity (no search involved) and robustness that characterize naive Bayes. We experimentally tested these approaches, using problems from the University of California at Irvine repository, and compared them to C4.5, naive Bayes, and wrapper methods for feature selection.
منابع مشابه
Induction of Selective Bayesian Network Classiiers
We present an algorithm for inducing Bayesian networks using feature selection. The algorithm selects a subset of attributes that maximizes predictive accuracy prior to the network learning phase, thereby incorporating a bias for small networks that retain high predictive accuracy. We compare the behavior of this selective Bayesian network classiier with that of (a) Bayesian network classiiers ...
متن کاملBayesian Classifiers Are Large Margin Hyperplanes in a Hilbert Space
Bayesian algorithms for Neural Networks are known to produce classiiers which are very resistant to overrtting. It is often claimed that one of the main distinctive features of Bayesian Learning Algorithms is that they don't simply output one hypothesis, but rather an entire distribution of probability over an hypothesis set: the Bayes posterior. An alternative perspective is that they output a...
متن کاملBiological Data Mining Using Bayesian Neural Networks: A Case Study
Biological data mining is the activity of nding signiicant information in biomolecular data. The signiicant information may refer to motifs, clusters, genes, and protein signatures. This paper presents an example of biological data mining: the recognition of promoters in DNA. We propose a two-level ensemble of classiiers to recognize E. Coli promoter sequences. The rst-level classiiers include ...
متن کاملRecognizing Promoters in DNA Using Bayesian Neural Networks
Binary data classiication is to recognize positive data from unlabeled test data that may contain both positive and negative data. In this paper we propose a two-level approach to recognize E. Coli promoters in unlabeled DNA containing both promoter and non-promoter sequences. The rst-level classiiers include three Bayesian neural networks which learn from three diierent feature sets. The outpu...
متن کاملBayesian Voting Schemes as Large MarginClassi
It is often claimed that one of the main distinctive features of Bayesian Learning Algorithms for neural networks is that they don't simply output one hypothesis, but rather an entire distribution of probability over an hypothesis set: the Bayes posterior. An alternative perspective is that they output a linear combination of classiiers, whose coeecients are given by Bayes theorem. This can be ...
متن کامل